Compiler-Directed Cache Assist Adaptivity

نویسندگان

  • Xiaomei Ji
  • Dan Nicolaescu
  • Alexander V. Veidenbaum
  • Alexandru Nicolau
  • Rajesh K. Gupta
چکیده

The performance of a traditional cache memory hierarchy can be improved by utilizing mechanisms such as a victim cache or a stream bu er (cache assists). The amount of on{ chip memory for cache assist is typically limited for technological reasons. In addition, the cache assist size is limited in order to maintain a fast access time. Performance gains from using a stream bu er or a victim cache, or a combination of the two, varies from program to program as well as within a program. Therefore, given a limited amount of cache assist memory, there is a need and a potential for \adaptivity" of the cache assists i.e., an ability to vary their relative size within the bounds of the cache assist memory size. We propose and study a compiler-driven adaptive cache assist organization and its e ect on system performance. Several adaptivity mechanisms are proposed and investigated. The results show that a cache assist that is adaptive at loop level clearly improves the cache memory performance, has low overhead, and can be easily implemented. 2

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Compilerrdirected Cache Assist Adaptivity ?

The performance of a traditional cache memory hierarchy can be improved by utilizing mechanisms such as a victim cache or a stream bu er (cache assists). The amount of on{chip memory for cache assist is typically limited for technological reasons. In addition, the cache assist size is limited in order to maintain a fast access time. Performance gains from using a stream bu er or a victim cache,...

متن کامل

Compiler-Directed Cache Line Size Adaptivity

The performance of a computer system is highly dependent on the performance of the cache memory system. The traditional cache memory system has an organization with a line size that is xed at design time. Miss rates for di erent applications can be improved if the line size could be adjusted dynamically at run time. We propose a system where the compiler can set the cache line size for di erent...

متن کامل

Multiprocessor Cache Coherence : The

The performance of large-scale shared-memory multiprocessors can be greatly improved if they can cache remote shared data in the private caches of the processors. However, maintaining cache coherence for such systems remains a challenge. Although hardware directory schemes give good performance, they might be too complicated and expensive for large-scale multiprocessors. This tutorial article p...

متن کامل

Eecient Integration of Compiler-directed Cache Coherence and Data Prefetching Compiler-directed Cache Coherence and Data Prefetching

Cache coherence enforcement and memory latency reduction and hiding are very important and challenging problems in the design of large-scale distributed shared-memory (DSM) multiprocessors. We propose an integrated approach to solve these problems through a compiler-directed cache coherence scheme called the Cache Coherence with Data Prefetching (CCDP) scheme. The CCDP scheme enforces cache coh...

متن کامل

Static Analysis of Parameterized Loop Nests for Energy EÆcient Use of Data Caches

Caches are an important part of architectural and compiler low-power strategies by reducing memory accesses and energy per access. In this paper, we examine eÆcient utilization of data caches for low power in an adaptive memory hierarchy. We focus on the optimization of data reuse through the static analysis of line size adaptivity. We present an approach that enables the quanti cation of data ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2000